In 2019 Kerdels and Peters proposed a grid cell model (GCM) based on a Differential Growing Neural Gas (DGNG) network architecture as a computationally efficient way to model an Autoassociative Memory Cell (AMC) \cite{Kerdels_Peters_2019}. An important feature of the DGNG architecture with respect to possible applications in the field of computational neuroscience is its \textit{capacity} refering to its capability to process and uniquely distinguish input signals and therefore obtain a valid representation of the input space. This study evaluates the capacity of a two layered DGNG grid cell model on the Fashion-MNIST dataset. The focus on the study lies on the variation of layer sizes to improve the understanding of capacity properties in relation to network parameters as well as its scaling properties. Additionally, parameter discussions and a plausability check with a pixel/segment variation method are provided. It is concluded, that the DGNG model is able to obtain a meaningful and plausible representation of the input space and to cope with the complexity of the Fashion-MNIST dataset even at moderate layer sizes.
translated by 谷歌翻译
We consider a sequential decision making task where we are not allowed to evaluate parameters that violate an a priori unknown (safety) constraint. A common approach is to place a Gaussian process prior on the unknown constraint and allow evaluations only in regions that are safe with high probability. Most current methods rely on a discretization of the domain and cannot be directly extended to the continuous case. Moreover, the way in which they exploit regularity assumptions about the constraint introduces an additional critical hyperparameter. In this paper, we propose an information-theoretic safe exploration criterion that directly exploits the GP posterior to identify the most informative safe parameters to evaluate. Our approach is naturally applicable to continuous domains and does not require additional hyperparameters. We theoretically analyze the method and show that we do not violate the safety constraint with high probability and that we explore by learning about the constraint up to arbitrary precision. Empirical evaluations demonstrate improved data-efficiency and scalability.
translated by 谷歌翻译
ICECUBE是一种用于检测1 GEV和1 PEV之间大气和天体中微子的光学传感器的立方公斤阵列,该阵列已部署1.45 km至2.45 km的南极的冰盖表面以下1.45 km至2.45 km。来自ICE探测器的事件的分类和重建在ICeCube数据分析中起着核心作用。重建和分类事件是一个挑战,这是由于探测器的几何形状,不均匀的散射和冰中光的吸收,并且低于100 GEV的光,每个事件产生的信号光子数量相对较少。为了应对这一挑战,可以将ICECUBE事件表示为点云图形,并将图形神经网络(GNN)作为分类和重建方法。 GNN能够将中微子事件与宇宙射线背景区分开,对不同的中微子事件类型进行分类,并重建沉积的能量,方向和相互作用顶点。基于仿真,我们提供了1-100 GEV能量范围的比较与当前ICECUBE分析中使用的当前最新最大似然技术,包括已知系统不确定性的影响。对于中微子事件分类,与当前的IceCube方法相比,GNN以固定的假阳性速率(FPR)提高了信号效率的18%。另外,GNN在固定信号效率下将FPR的降低超过8(低于半百分比)。对于能源,方向和相互作用顶点的重建,与当前最大似然技术相比,分辨率平均提高了13%-20%。当在GPU上运行时,GNN能够以几乎是2.7 kHz的中位数ICECUBE触发速率的速率处理ICECUBE事件,这打开了在在线搜索瞬态事件中使用低能量中微子的可能性。
translated by 谷歌翻译
路径计划是设计机器人行为的关键算法方法。基于抽样的方法,例如快速探索随机树(RRT)或概率路线图,是针对路径计划问题的突出算法解决方案。尽管其指数收敛速率,RRT只能找到次优路径。另一方面,$ \ textrm {rrt}^*$是RRT广泛​​使用的扩展名,保证了寻找最佳路径的概率完整性,但在复杂环境中缓慢收敛而在实践中遭受痛苦。此外,现实世界中的机器人环境通常是可观察到的,或者描述的动力学不好,施放了$ \ textrm {rrt}^*$在复杂任务中的应用。本文研究了用于机器人路径计划的流行蒙特卡洛树搜索(MCTS)算法的新型算法公式。值得注意的是,我们通过分析和证明其指数的收敛速率(MCPP)在完全可观察到的马尔可夫决策过程(MDP)的一部分中,并证明其指数收敛速率,而另一部分则是其概率的完整性假设有限的距离可观察性(证明草图),在部分可观察的MDP(POMDP)中找到可行的路径。我们的算法贡献使我们能够采用最近提出的MCT的变体,并具有不同的勘探策略来进行机器人路径计划。我们在模拟的2D和3D环境中进行了7度自由度(DOF)操纵器以及现实世界机器人路径计划任务中的实验评估,证明了MCPP在POMDP任务中的优势。
translated by 谷歌翻译
事实证明,基于文本的模拟环境是机器学习方法的有效测试床。可负担提取的过程可用于在这种环境中生成可能的互动操作。在本文中,研究了在负担能力提取过程中利用外部知识数据库(特别是概念网)的功能和挑战。在交互式小说(IF)平台Textworld和Jericho上引入和评估了一种用于自动负担提取的算法。为此,将收集的负担转换为IF代理的文本命令。为了探究自动评估过程的质量,进行了另外的人类基线研究。该论文说明,尽管有一些挑战,但原则上可以将外部数据库用于负担得出。本文以进一步修改和改进过程的建议结束。
translated by 谷歌翻译
尽管近期因因果推断领域的进展,迄今为止没有关于从观察数据的收集治疗效应估算的方法。对临床实践的结果是,当缺乏随机试验的结果时,没有指导在真实情景中似乎有效的指导。本文提出了一种务实的方法,以获得从观察性研究的治疗效果的初步但稳健地估算,为前线临床医生提供对其治疗策略的信心程度。我们的研究设计适用于一个公开问题,估算Covid-19密集护理患者的拳击机动的治疗效果。
translated by 谷歌翻译
While the capabilities of autonomous systems have been steadily improving in recent years, these systems still struggle to rapidly explore previously unknown environments without the aid of GPS-assisted navigation. The DARPA Subterranean (SubT) Challenge aimed to fast track the development of autonomous exploration systems by evaluating their performance in real-world underground search-and-rescue scenarios. Subterranean environments present a plethora of challenges for robotic systems, such as limited communications, complex topology, visually-degraded sensing, and harsh terrain. The presented solution enables long-term autonomy with minimal human supervision by combining a powerful and independent single-agent autonomy stack, with higher level mission management operating over a flexible mesh network. The autonomy suite deployed on quadruped and wheeled robots was fully independent, freeing the human supervision to loosely supervise the mission and make high-impact strategic decisions. We also discuss lessons learned from fielding our system at the SubT Final Event, relating to vehicle versatility, system adaptability, and re-configurable communications.
translated by 谷歌翻译
Attention mechanisms form a core component of several successful deep learning architectures, and are based on one key idea: ''The output depends only on a small (but unknown) segment of the input.'' In several practical applications like image captioning and language translation, this is mostly true. In trained models with an attention mechanism, the outputs of an intermediate module that encodes the segment of input responsible for the output is often used as a way to peek into the `reasoning` of the network. We make such a notion more precise for a variant of the classification problem that we term selective dependence classification (SDC) when used with attention model architectures. Under such a setting, we demonstrate various error modes where an attention model can be accurate but fail to be interpretable, and show that such models do occur as a result of training. We illustrate various situations that can accentuate and mitigate this behaviour. Finally, we use our objective definition of interpretability for SDC tasks to evaluate a few attention model learning algorithms designed to encourage sparsity and demonstrate that these algorithms help improve interpretability.
translated by 谷歌翻译
Recent advances in deep learning have enabled us to address the curse of dimensionality (COD) by solving problems in higher dimensions. A subset of such approaches of addressing the COD has led us to solving high-dimensional PDEs. This has resulted in opening doors to solving a variety of real-world problems ranging from mathematical finance to stochastic control for industrial applications. Although feasible, these deep learning methods are still constrained by training time and memory. Tackling these shortcomings, Tensor Neural Networks (TNN) demonstrate that they can provide significant parameter savings while attaining the same accuracy as compared to the classical Dense Neural Network (DNN). In addition, we also show how TNN can be trained faster than DNN for the same accuracy. Besides TNN, we also introduce Tensor Network Initializer (TNN Init), a weight initialization scheme that leads to faster convergence with smaller variance for an equivalent parameter count as compared to a DNN. We benchmark TNN and TNN Init by applying them to solve the parabolic PDE associated with the Heston model, which is widely used in financial pricing theory.
translated by 谷歌翻译
Artificial neural networks can learn complex, salient data features to achieve a given task. On the opposite end of the spectrum, mathematically grounded methods such as topological data analysis allow users to design analysis pipelines fully aware of data constraints and symmetries. We introduce a class of persistence-based neural network layers. Persistence-based layers allow the users to easily inject knowledge about symmetries (equivariance) respected by the data, are equipped with learnable weights, and can be composed with state-of-the-art neural architectures.
translated by 谷歌翻译